Recently, the research team led by Prof. WANG Hongqiang from the Hefei Institutes of Physical Science of the Chinese Academy of Sciences proposed a wide-ranging cross-modality machine vision AI model.
This model overcame the limitations of traditional single-domain models in handling cross-modality information and achieved new breakthroughs in cross-modality image retrieval technology.
Cross-modality machine vision is a major challenge in AI, as it involves finding consistency and complementarity between different types of data. Traditional methods focus on images and features but are limited by issues like information granularity and lack of data. Compared to traditional methods, researchers found that detail associations are more effective in maintaining consistency across modalities.
In this study, the team introduced a Wide-Ranging Information Mining Network (WRIM-Net). This model created global region interactions to extract detailed associations across various domains, such as spatial, channel, and scale domains, emphasizing modality invariant information mining across a broad range.
Additionally, the research team guided the network to effectively extract modality-invariant information by designing a cross-modality key-instance contrastive loss. Experimental validation showed the model's effectiveness on both standard and large-scale cross-modality datasets, achieving over 90% in several key performance metrics for the first time.
This model can be applied in various fields of artificial intelligence, including visual traceability and retrieval as well as medical image analysis, according to the team.
Cross-modality image retrieval workflow based on the Model (Image by WANG Hongqiang)
86-10-68597521 (day)
86-10-68597289 (night)
86-10-68511095 (day)
86-10-68512458 (night)
cas_en@cas.cn
52 Sanlihe Rd., Xicheng District,
Beijing, China (100864)